Bernoulli numbers turn up in a surprising number of contexts. This page will document the most outstanding and make sense of why these numbers appear therein.
Definition of Bernoulli Numbers
Trigonometric Power Series
Sums of Products of Bernoulli Numbers
Sums of Powers of Integers
Bernoulli Polynomials
Euler-Maclaurin Summation
Relations to Zeta Functions
These numbers are defined by the generating function
The sum on the left-hand side is the inverse of an exponential series with powers shifted down by one. Rearranging the definition gives
The first term on the right-hand side is constant, so that . The remaining powers must then have coefficients equal to zero. Picking out an arbitrary power gives
This summation that equals zero is all important in detecting the presence of Bernoulli numbers, and will be used time and again in what follows.
Since binomial coefficients are positive by definition, some of the Bernoulli numbers must be negative for the total sum to equal zero. The numbers can be evaluated recursively using
With the aid of a CAS or a bit of calculational persistence, the first few values are
All odd numbers after are identically zero. This is easy to show by separating the first two terms of the sum:
Since the final right-hand side is an even function, there can be no odd terms on the left-hand side. The coefficients of these terms are necessarily zero.
Even Bernoulli numbers alternate in sign, but this will not be important in what follows.
Because of the generating function definition, Bernoulli numbers appear prominently in the power series of trigonometric functions other than sine and cosine. Consider first the hyperbolic cotangent:
The index of the sum can be rewritten in the last step because all the odd Bernoulli numbers in the sum are identically zero. The singular term can be combined with the sum but is kept separate for clarity.
For the circular cotangent, let and, noting that , multiply by the i from the denominator of the left-hand side:
The hyperbolic tangent has a positive sign in its denominator, requiring a change of sign in the denominator of the the generating function. First note the partial fraction decomposition
so that one has
The power series for the hyperbolic tangent is thus
where again the index of the sum can be rewritten in the last step because all the odd Bernoulli numbers in the sum are identically zero.
For the circular tangent, let and divide by the i from the numerator of the left-hand side:
For the hyperbolic cosecant, use an intermediate partial fraction decomposition:
One more time the index of the sum can be rewritten in the last step because all the odd Bernoulli numbers in the sum are identically zero. The singular term is again kept separate for clarity.
For the circular cosecant, let and multiply by the i from the denominator of the left-hand side:
Series for the hyperbolic and circular secants cannot be written in terms of Bernoulli numbers, but rather define a new set known as Euler numbers.
Consider the consistency of power series in the context of the two derivatives
Including singular terms in sums for conciseness, these two relations can be expressed as
Picking out an arbitrary power of the independent variable on each side of the equation, these two expressions are equivalent to
This will appear much simpler with the abbreviations
Note that both sums exhibit an interchange symmetry between and . Keeping this in mind, the two equalities can be written
A bit of algebra establishes that , except for when the system is indeterminate. From either of these two relations, and for , one then has
This is somewhat surprising: the extra power of two inside the second summation has no effect on the result. These summations can also be written without the first and last terms as
This result has been known since Euler, but perhaps this is a novel derivation.
Bernoulli numbers first arose in the context of summing powers of integers. To motivate the general form, first consider a trivial sum over zeroth powers:
The corresponding sum over first powers uses a trick of Gauss,
where the first and last terms are added and multiplied by half the number of such intermediate sums. This form remains an integer for n both even and odd.
The sum over squares of integers is most easily done using a trick of Knuth. Consider an identity for the sum of cubes:
The sum over cubes cancels on both sides, but one can then solve for the sum of squares using previous results:
This method can be used for all subsequent sums, as well as the previous two results.
These examples are enough to see that the final result begins with a power one more than that in the sum, descending to a single power. Thus assume that the sum over integers to an arbitrary power has the form
For p = 0, this form is satisfied with c1 = 1. Consider induction on the assumed form:
Before expanding binomials, let in the second line to simplify,
so that now the right-hand side can be easily eliminated:
The indices k and m describe a triangular region in their coordinate plane. Diagonals in this triangular region correspond to the difference between the two indices. These differences run from zero, when the two indices are equal, up to the power p, when the indices differ the most. Working down in value from the uppermost diagonal, when the indices are equal, produces an new upper limit on the second index m. The double sum can thus be rearranged as
One can now set coefficients of powers of n equal to zero for coefficient values. The highest power gives
while the next highest power gives
which is curiously independent of the power p. Both of these results are consistent with the examples at the beginning of this section. For the general case one has
and the result just above suggest one should look at a descending index defined by . The general case then becomes
where a possible overall minus sign is ignored as inconsequential. Now consider casting this in the form of the all-important summation for Bernoulli numbers:
One would think at first that the alternating nature of the sum in question would make this impossible, except for one fact: all odd Bernoulli numbers are zero except . Apart from explicitly adjusting the sign of the coefficient containing this value, one can ignore the alternating sign in the series.
One thus needs to have
Ignoring factors not dependent of the index of summation k, this gives
Renaming the index and adjusting the normalization factor to match the first two values gives the result
The extra minus sign accounts for the one term in the alternating series that contains the nonzero odd Bernoulli number .
The final form of the sum over integers to an arbitrary power is thus
and the leading terms are
where as noted , along every other subsequent coefficient, due to odd Bernoulli numbers being zero after the first. The coefficients here are naturally consistent with the examples at the beginning of this section. The linear term will not appear if multiplied by an odd Bernoulli number other than .
This form of the sum has a rather surprising implication,
since only one term of the sum is modified. That is, a sum over one less integer can be written with the standard Bernoulli numbers without modification.
These polynomials are defined by the generating function
Inserting the definition of the Bernoulli numbers gives
where the last rearrangement picks out individual powers of the variable t. The polynomials are thus given in terms of the numbers by
Again with the aid of a CAS or a bit of calculational persistence, the first few polynomials are
From the definition it is trivial to see that
The value of the polynomials at unity is in most cases identical, which can be seen by shifting the quantity n in the all-important summation for Bernoulli numbers:
This relation does not hold for
because the intermediate sum is nonzero: instead one has
The all-important summation leads to the following property of the polynomials for :
The derivative of the polynomials is
This relation holds for all values of the index and represents an Appell sequence. The relation can also be written in terms of an integral as
with a shift in index. Constants of integration are here chosen proportional to Bernoulli numbers for consistency. Evaluating the left-hand side as a definite integral from zero to one is an alternate way to show that the polynomials have identical values at these two endpoints, apart from the exception noted above.
The formula derived here provides a way to evaluate the error between an integral and its approximation as a sum. While the derivation given in venerable Whittaker & Watson is accurate, it provides no hint as to why Bernoulli polynomials are employed as opposed to any other set of similarly defined polynomials. This derivation will be clearer in that crucial aspect.
An integral over a small region can be approximated to first order by a trapezoid,
where the heights at the endpoints are averaged and multiplied by the base. The error in this first approximation is clearly the difference of the two sides of this equation:
Consider integration by parts of the following integral:
Setting the constant equal to the average of the two endpoints, this is identical to the expression for error, which can thus be represented
Now make the change of variable
so that the integral becomes
where the argument of the function is left as is for conciseness.
The procedure going forward is to perform repeated integration by parts on this error integral to produce a sum of decreasing error terms. The polynomials that will appear during this process constitute an Appell sequence.
Most derivations at this point recognize the first factor in the integrand as the Bernoulli polynomial
One way to do this is to focus on a final result that will be useful in evaluation. So far one has the exact statement
over a small region. For extended regions one needs to add the contributions from each small region,
where and . The first two terms represent the trapezoidal rule for estimating integrals: internal points contribute twice that of endpoints from adjacent small regions.
The third term also has contributions from adjacent regions, but given that it consists of integrals the contributions appear with opposite signs. If they can be made to cancel, the sum of error terms would then only depend on endpoint contributions. This will be the criterion for determining constants of integration.
Define an Appell-like sequence of polynomials with
The integration by parts of a general error integral can then be written
Clearly the condition that allows cancellation of interior error contributions is
If this condition holds, the estimate to the integral becomes
after repeated integration by parts. The remainder term R is simply the next error integral in the sequence and will not be considered further.
And now for the crucial part: determining the polynomials themselves. Starting from the one appearing in the integrand, the first few are
from which one can deduce the general form
Equating the value of this polynomials and zero and one gives the equation
With the choice this becomes
which is the all-important summation for Bernoulli numbers! Simply requiring these polynomials to have equal values at zero and one is enough to produce Bernoulli numbers and confirm that Bernoulli polynomials are indeed appropriate for the process. Comparing the general form here to the definition of Bernoulli polynomials gives
The approximation to the integral is thus
and since the odd Bernoulli numbers in the sum are zero, this can be written a bit more simply as
The sum here is in general asymptotic, requiring an upper cutoff to prevent divergence.
It is instructive to compare this derivation to that originally given by MacLaurin as described in this paper, but with modern notation replacing fluxions. Start with the Taylor series of a function at some initial point:
Integrating this from the initial point to one separated by unity will define a sum
whose first term is . Define a sequence of such sums with
where subscript on the sum indicates that its first term is : initial constant terms cancel in the difference between endpoints, leaving only terms corresponding to the upper endpoint.
It is not hard to see from writing out a few terms in each sum that a particular derivative will appear only in a finite number of sums. Combining the sums with coefficients to be determined and picking out individual derivatives on the right-hand sides gives
One can then solve for by choosing the coefficients to make remaining sums on the right-hand side zero. If this is done, then one can write
Summing over n unit steps to a final point, one can write the integral as
And how should the coefficients be chosen? With the remaining sums are all
which as anticipated is the all-important summation for Bernoulli numbers. The integral is thus
which as expected is equivalent to the result of the first derivation with .
While it may appear at first that Bernoulli numbers arise here in some sense from inverting a Taylor series, one can start the solution for derivatives at any arbitrary sum for the general result
Bernoulli numbers thus appear here, not from inverting a Taylor series per se, but from inverting its factorial coefficients. Curious...
Begin with the definition of the gamma function as an integral:
Change the variable of integration using ,
rearrange slightly and and sum over the introduced index,
where the intermediate step uses an infinite geometric series. The left-hand sum is the Hurwitz zeta function
This integral converges as long as to avoid the singularity at the origin. For general values of the independent variable, interpret this integral as part of a complex contour that comes in from positive infinity, circles the origin and returns to positive infinity. That is, consider the contour integral
The direct variable u has a complex argument of zero on the positive real axis. A contour that comes in from negative infinity has an argument of below the negative real axis, and an argument of upon returning to negative infinity above the negative real axis. When this contour is rotated to handle the negative variable these arguments are inverted, with the negative value on the part of the contour coming in from positive infinity above the positive real axis, and the positive value upon returning to positive infinity below the positive real axis.
Around the origin parametrize the contour as . On this part of the contour the integral becomes
where the left-hand side is exact, and approaches the right-hand value in the indicated limit. There is thus no contribution from this part of the contour, and the complete contour integral can be written
where explicit values of the phase have been separated out in the second step as per the description above of the contour. Combining intermediate results gives
using a reflection formula for the gamma function. One can now change the variable of integration using
where the contour is now rotated to the negative real axis, but still circles the origin. This representation is given here in a standard online reference.
Now insert the definition of Bernoulli polynomials,
and let argument z be equal to a negative integer. Under the residue theorem, only terms with
which is here in the standard online reference. Renaming the index gives the general relation
Since the Hurwitz zeta function is analytic in both parameters, this relation can be used to extend Bernoulli polynomials as analytic functions of both index and argument.
When the second argument of the Hurwitz zeta function is unity, it becomes the Riemann zeta function. Since the value of Bernoulli polynomials at unity is usually the corresponding Bernoulli number, one has immediately
This expression can appear with an additional alternating sign, which is unnecessary because the Bernoulli numbers that would be affected by such a sign are all zero.
Uploaded 2023.08.17 — Updated 2023.11.14 analyticphysics.com